Variational Network Quantization

نویسندگان

  • Jan Achterhold
  • Jan M. Köhler
  • Anke Schmeink
  • Tim Genewein
چکیده

In this paper, the preparation of a neural network for pruning and few-bit quantization is formulated as a variational inference problem. To this end, a quantizing prior that leads to a multi-modal, sparse posterior distribution over weights, is introduced and a differentiable Kullback-Leibler divergence approximation for this prior is derived. After training with Variational Network Quantization, weights can be replaced by deterministic quantization values with small to negligible loss of task accuracy (including pruning by setting weights to 0). The method does not require fine-tuning after quantization. Results are shown for ternary quantization on LeNet-5 (MNIST) and DenseNet (CIFAR-10).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Quantization of Thermodynamics, Supersecondary Quantization, and a New Variational Principle

In solving the problem of finding a temperature distribution which, at zero temperature, corresponds to superfluidity, i.e., to nonzero energy, the author tried to quantize free energy. This was done on the basis of supersecondary quantization whose special case is the usual secondary quantization for bosons and with the help of which new representations of the Schrödinger equation were obtaine...

متن کامل

Hölder continuity of a parametric variational inequality

‎In this paper‎, ‎we study the Hölder continuity of solution mapping to a parametric variational inequality‎. ‎At first‎, ‎recalling a real-valued gap function of the problem‎, ‎we discuss the Lipschitz continuity of the gap function‎. ‎Then under the strong monotonicity‎, ‎we establish the Hölder continuity of the single-valued solution mapping for the problem‎. ‎Finally‎, ‎we apply these resu...

متن کامل

INTERNATIONAL CENTRE FOR THEORETICAL PHYSICS Dissipative Quantum Mechanics: The Generalization of the Canonical Quantization and von Neumann Equation

– Sedov variational principle, which is the generalization of the least action principle for the dissipative processes is used to generalize the canoni-cal quantization and von Neumann equation for dissipative systems (particles and strings). PACS 03.65 – Quantum theory; quantum mechanics PACS 05.30 – Quantum statistical mechanics

متن کامل

Discrete-valued Neural Networks Using Variational Inference

The increasing demand for neural networks (NNs) being employed on embedded devices has led to plenty of research investigating methods for training low precision NNs. While most methods involve a quantization step, we propose a principled Bayesian approach where we first infer a distribution over a discrete weight space from which we subsequently derive hardware-friendly low precision NNs. To t...

متن کامل

Discrete-valued Neural Networks Using Variational Inference

The increasing demand for neural networks (NNs) being employed on embedded devices has led to plenty of research investigating methods for training low precision NNs. While most methods involve a quantization step, we propose a principled Bayesian approach where we first infer a distribution over a discrete weight space from which we subsequently derive hardware-friendly low precision NNs. To t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2018